---
title: Import to DataRobot directly
description: This section provides detailed steps for importing without using the catalog, by drag and drop, URL, and HDFS.

---

# Import to DataRobot directly {: #import-data-to-DataRobot-directly }

This section describes detailed steps for importing data to DataRobot. Before you import data, review DataRobot's [data guidelines](file-types) to understand dataset requirements, including file formats and sizes.

!!! notes
    This section assumes that your data sources are configured. If not, see the [JDBC connection](data-conn) instructions on selecting a data connection and data source, as well as creating SQL queries.

## To get started {: #to-get-started }

The first step to building models is to import your data. To get started:

1. Create a new DataRobot project in either of the following ways:

	* Sign in to DataRobot and click the DataRobot logo in the upper left corner.

	* Open the [**Projects** folder](manage-projects#create-a-new-project) in the upper right corner and click the **Create New Project** link.

2. Once the new project page is open, [select a method to import](#import-methods) an acceptable file type to the page. (Accepted types are listed at the bottom of the screen.) If a particular upload method is disabled on your cluster, the corresponding ingest button will be grayed out.


## Import methods {: #import-methods }

Once you sign in to DataRobot, you can import data and start a project. The following are ways to import your data.

!!! note
    Some import methods need to be configured before users in your organization can use them, as noted in the following sections.


| Method | Description |
|---|---|
| [Drag and drop](#drag-and-drop) | Drag a dataset into DataRobot to begin an upload. |
| [Use an existing data source](#use-an-existing-data-source) | Import from a configured data source. |
| [Import a dataset from a URL](#import-a-dataset-from-a-url) | Specify a URL from which to import data. |
| [Import local files](#import-local-files) | Browse to a local file and import. |
| [Import files from S3](#import-files-from-s3) | Upload from an AWS S3 bucket. |
| [Import files from Google Cloud Storage](#import-files-from-google-cloud-storage) | Import directly from Google Cloud. |
| [Import files from Azure Blob Storage](#import-files-from-azure-blob-storage) | Import directly from Azure Blob. |

!!! note
    A particular upload method may be disabled on your cluster, in which case a button for that method does not appear. Contact your system administrator for information about the configured import methods.

For larger datasets, DataRobot provides [special handling](fast-eda) that lets you see your data earlier and select project options earlier.

### Drag and drop {: #drag-and-drop }

To use drag and drop, simply drag a file onto the app. Note, however, that when dropping large files (greater than 100MB) the upload process may hang. If that happens:

* Try again.

* Compress the file into a supported [format](file-types#data-formats) and then try again.

* Save the file to a remote data store (e.g., S3) and use URL ingest, which is more reliable for large files.

* If security is a concern, use a temporarily signed S3 URL.

### Use an existing data source {: #use-an-existing-data-source }

You can use this method if you have already configured data sources. If not, see the [JDBC](data-conn) connection instructions for details on selecting a data connection and data source, as well as creating SQL queries.

!!! note
    When DataRobot ingests from the data source option, it makes a copy of the selected database rows for your use in the project.

To use an existing data source:

1. On the new project screen, click **Data Source**.

2. Select the desired data source and click **Next**.

3. Use saved credentials or enter new credentials for the database configured.

4. Click **Save and sign in**.


###  Import a dataset from a URL {: #import-a-dataset-from-a-url }

To import data from a URL:

1. On the new project screen, click **URL**.

2. Enter the URL of the dataset. It can be [local](#import-local-files), HTTP, HTTPS, [Google Cloud Storage](#import-files-from-google-cloud-storage), [Azure Blob Storage](#import-files-from-azure-blob-storage), or [S3](#import-files-from-s3).

    !!! note
        The ability to import from Google Cloud, Azure Blob Storage, or S3 using a URL needs to be configured for your organization's installation.

4. Click **Create New Project** to create a new project.


### Import local files {: #import-local-files }

Instead of copying data to the client and then uploading it via the browser, you can specify the URL link as <code>file:///local/file/location</code>. DataRobot will then ingest the file from the network storage drive connected to the cluster. This import method needs to be configured for your organization's installation.

!!! note
    The ability to load locally mounted files directly into DataRobot is not available for managed AI Platform users.

### Import files from S3 {: #import-files-from-s3 }

Self-Managed AI Platform installations with this import method configured can ingest S3 files via a URL by specifying the link to S3 as `s3://<bucket-name>/<file-name.csv>` (instead of, for example, `https://s3.amazonaws.com/bucket/file?AWSAccessKeyId...`). This allows you to ingest files from S3 without setting your object and buckets to public.

!!! note
    This method is disabled for managed AI Platform users. Instead, import S3 files using one of the following methods:

    * Using an Amazon S3 [data connection](data-conn).
    * Generate a pre-signed URL allowing public access to S3 buckets with authentication, then you can use a direct URL to ingest the dataset.

### Import files from Google Cloud Storage {: #import-files-from-google-cloud-storage }

You can configure DataRobot to directly import files stored in Google Cloud Storage using the link `gs://<bucket-name>/<file-name.csv>`. This import method needs to be configured for your organization's installation.

!!! note
    The ability to import files using the `gs://<bucket-name>/<file-name.csv>` link is not available for managed AI Platform users.

### Import files from Azure Blob Storage {: #import-files-from-azure-blob-storage }

It is possible to directly import files stored in Azure Blob Storage using the link `azure_blob://<container-name>/<file-name.csv>`. This import method needs to be configured for your organization's installation.

!!! note
    The ability to import files using the `azure_blob://<container-name>/<file-name.csv>` link is not available for managed AI Platform users.

## Project creation and analysis {: #project-creation-and-analysis }

After you select a data source and import your data, DataRobot creates a new project.  This first [*exploratory data analysis*](eda-explained) step is known as *EDA1*. (See the section on ["Fast EDA"](fast-eda) to understand how DataRobot handles larger datasets.)

Progress messages indicate that the file is being processed.

![](images/data-upload-in-progress.png)

When [EDA1](eda-explained#eda1) completes, DataRobot displays the **Start** screen. From here you can scroll down or click the **Explore** link to view a data summary. You can also [specify the target feature](model-data#set-the-target-feature) to use for predictions.

![](images/eda1.png)

Once you're in the data section, you can:

![](images/eda1-data.png)

* Click **View Raw Data** (1) to display a modal presenting up to a 1MB random sample of the raw data table DataRobot will be building models with:

	![](images/raw-data.png)

* [Set your target](model-data#set-the-target-feature) (2) by mousing over a feature name in the data display.

* Work with [feature lists](feature-lists) (3).

You can also [view a histogram](histogram) for each feature. The histogram provides several options for modifying the display to help explain the feature and its relationship to the dataset.

More information becomes available once you set a target feature and begin your [model build](model-data), which is the next step.
